An Overview of Distance Metric Learning

نویسنده

  • Liu Yang
چکیده

In our previous comprehensive survey [41], we have categorized the disparate issues in distance metric learning. Within each of the four categories, we have summarized existing work, disclosed their essential connections, strengths and weaknesses. The first category is supervised distance metric learning, which contains supervised global distance metric learning, local adaptive supervised distance metric learning, Neighborhood Component Analysis (NCA) [13], and Relevant Components Analysis (RCA) [1]. The second category is unsupervised distance metric learning, covering linear (Principal Component Analysis (PCA) [14], Multidimensional Scaling (MDS) [5]) and nonlinear embedding methods (ISOMAP [35], Locally Linear Embedding (LLE) [30], and Laplacian Eigenamp (LE) [2]). We further unify these algorithms into a common framework based on the embedding computation. The third category, which is maximum margin based distance metric learning approaches, includes the large margin nearest neighbor based distance metric learning methods and semi-definite Programming (SDP) methods to solve the kernelized margin maximization problem. And the fourth category discussing kernel methods towards learning distance metrics, covers kernel alignment [28] and its SDP approaches [26], and also the extension work of learning the idealized kernel [25]. In addition to this survey [41], here we provide a complete and updated summarization of the related work on both unsupervised distance metric learning and supervised distance metric learning, including the most recent work in the area of distance metric learning. Many unsupervised distance metric learning algorithms are essentially for the purpose of unsupervised dimensionality reduction, i.e. learning a low-dimensional embedding of the original feature space. This group of methods can be divided into nonlinear and linear methods. The well known algorithms for nonlinear unsupervised dimensionality reduction are ISOMAP [35], Locally Linear Embedding (LLE) [30], and Laplacian Eigenamp (LE) [2]. ISOMAP seeks the subspace that best preserves the geodesic distances between any two data points, while LLE and LE focus on the preservation of local neighbor structure. An improved and stable version of LLE is achieved in [45], by introducing multiple linearly independent local weight vectors for each neighborhood. Among the linear methods, Principal Component Analysis (PCA) [14] finds the subspace that best preserves the variance of the data; Multidimensional Scaling (MDS) [5] finds the low-rank projection that best preserves the inter-point distance given by the pairwise distance matrix; Independent components analysis (ICA) [4] seeks a linear transformation to coordinates in which the data are maximally statistically indepen-

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

یادگیری نیمه نظارتی کرنل مرکب با استفاده از تکنیک‌های یادگیری معیار فاصله

Distance metric has a key role in many machine learning and computer vision algorithms so that choosing an appropriate distance metric has a direct effect on the performance of such algorithms. Recently, distance metric learning using labeled data or other available supervisory information has become a very active research area in machine learning applications. Studies in this area have shown t...

متن کامل

An Effective Approach for Robust Metric Learning in the Presence of Label Noise

Many algorithms in machine learning, pattern recognition, and data mining are based on a similarity/distance measure. For example, the kNN classifier and clustering algorithms such as k-means require a similarity/distance function. Also, in Content-Based Information Retrieval (CBIR) systems, we need to rank the retrieved objects based on the similarity to the query. As generic measures such as ...

متن کامل

Composite Kernel Optimization in Semi-Supervised Metric

Machine-learning solutions to classification, clustering and matching problems critically depend on the adopted metric, which in the past was selected heuristically. In the last decade, it has been demonstrated that an appropriate metric can be learnt from data, resulting in superior performance as compared with traditional metrics. This has recently stimulated a considerable interest in the to...

متن کامل

A Survey on Metric Learning for Feature Vectors and Structured Data

The need for appropriate ways to measure the distance or similarity between data is ubiquitous in machine learning, pattern recognition and data mining, but handcrafting such good metrics for specific problems is generally difficult. This has led to the emergence of metric learning, which aims at automatically learning a metric from data and has attracted a lot of interest in machine learning a...

متن کامل

Algebraic distance in algebraic cone metric spaces and its properties

In this paper, we prove some properties of algebraic cone metric spaces and introduce the notion of algebraic distance in an algebraic cone metric space. As an application, we obtain some famous fixed point results in the framework of this algebraic distance.

متن کامل

The Banach Type Contraction for Mappings on Algebraic Cone Metric Spaces Associated with An Algebraic Distance and Endowed with a Graph

In this work, we define the notion of an algebraic distance in algebraic cone metric spaces defined by Niknam et al. [A. Niknam, S. Shamsi Gamchi and M. Janfada, Some results on TVS-cone normed spaces and algebraic cone metric spaces, Iranian J. Math. Sci. Infor. 9 (1) (2014), 71--80] and introduce some its elementary properties. Then we prove the existence and uniqueness of fixed point for a B...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007